Nonlinear problems in analysis of Krylov subspace methods
نویسنده
چکیده
Consider a system of linear algebraic equations Ax = b where A is an n by n real matrix and b a real vector of length n. Unlike in the linear iterative methods based on the idea of splitting of A, the Krylov subspace methods, which are used in computational kernels of various optimization techniques, look for some optimal approximate solution xn in the subspaces Kn(A, b) = span{b, Ab, . . . , A b}, n = 1, 2, . . . (here we assume, with no loss of generality, x = 0). As a consequence, though the problem Ax = b is linear, Krylov subspace methods are not. Their convergence behaviour cannot be viewed as an (unimportant) initial transient stage followed by the subsequent convergence stage. Apart from very simple, and from the point of view of Krylov subspace methods uninteresting cases, it cannot be meaningfully characterized by an asymptotic rate of convergence. In Krylov subspace methods such as the conjugate gradient method (CG) or the generalized minimal residual method (GMRES), the optimality at each step over Krylov subspaces of increasing dimensionality makes any linearized description inadequate. CG applied to Ax = b with a symmetric positive definite A can be viewed as a method for numerical minimization the quadratic functional 1/2(Ax, x) − (b, x). In order to reveal its nonlinear character, we consider CG a matrix formulation of the Gauss-Christoffel quadrature, and show that it essentially solves the classical Stieltjes moment problem. Moreover, though the CG behaviour is fully determined by the spectral decomposition of the problem, the relationship between convergence and spectral information is nothing but simple. We will explain several phenomena where an intuitive commonly used argumentation can lead to wrong conclusions, which can be found in the literature. We also show that rounding error analysis of CG brings fundamental understanding of seemingly unrelated problems in convergence analysis and in theory of the Gauss-Christoffel quadrature [5, 3, 4]. In remaining time we demonstrate that in the unsymmetric case the spectral information is not generally sufficient for description of behaviour of Krylov subspace methods. In particular, given an arbitrary prescribed convergence history of GMRES and an arbitrary prescribed spectrum of the system matrix, there is always a system Ax = b such that GMRES follows the prescribed convergence while A has the prescribed spectrum [2, 1].
منابع مشابه
Convergence analysis of Krylov subspace methods †
One of the most powerful tools for solving large and sparse systems of linear algebraic equations is a class of iterative methods called Krylov subspace methods. Their significant advantages like low memory requirements and good approximation properties make them very popular, and they are widely used in applications throughout science and engineering. The use of the Krylov subspaces in iterati...
متن کاملPrediction of critical boundaries in two-parameter nonlinear eigenvalue problems
Prediction of critical boundaries (i.e. curves of singular points) in two-parameter nonlinear eigenvalue problems is considered. A generalization of the Krylov subspace for the one-parameter linear eigenvalue problem to the two-parameter case, suitable for the prediction of critical boundaries, is introduced. Methods for the eecient computation of the solution surface are developed. A case of t...
متن کاملPreconditioned Krylov Subspace Methods in Nonlinear Optimization
One of the possible ways of solving general problems of constrained nonlinear optimization is to convert them into a sequence of unconstrained problems. Then the need arises to solve an unconstrained optimization problem reliably and efficiently. For this aim, Newton methods are usually applied, often in combination with sparse Cholesky decomposition. In practice, however, this approach may not...
متن کاملKrylov Methods for Nonlinear Eigenvalue Problems
We present two generalisations of the Krylov subspace method, Arnoldi for the purpose of applying them to nite dimensional eigenvalue problems nonlinear in the eigenvalue parameter. The rst method is called nonlinear rational Krylov subspace and approximates and updates the projection of a linearised problem by nesting a one-sided secant method with Arnoldi. The second method, called nonlinear ...
متن کاملTensor-Krylov methods for large nonlinear equations
In this paper, we describe tensor methods for large systems of nonlinear equations based on Krylov subspace techniques for approximately solving the linear systems that are required in each tensor iteration. We refer to a method in this class as a tensor-Krylov algorithm. We describe comparative testing for a tensor-Krylov implementation versus an analogous implementation based on a Newton-Kryl...
متن کامل